11 research outputs found
3D Time-Based Aural Data Representation Using D4 Library’s Layer Based Amplitude Panning Algorithm
Presented at the 22nd International Conference on Auditory Display (ICAD-2016)The following paper introduces a new Layer Based Amplitude Panning algorithm and supporting D4 library of rapid prototyping tools for the 3D time-based data representation using sound. The algorithm is designed to scale and support a broad array of configurations, with particular focus on High Density Loudspeaker Arrays (HDLAs). The supporting rapid prototyping tools are designed to leverage oculocentric strategies to importing, editing, and rendering data, offering an array of innovative approaches to spatial data editing and representation through the use of sound in HDLA scenarios. The ensuing D4 ecosystem aims to address the shortcomings of existing approaches to spatial aural representation of data, offers unique opportunities for furthering research in the spatial
data audification and sonification, as well as transportable and scalable spatial media creation and production
Sonifying 2D Cellular Behavior Using Cellular Stehoscope
Presented at the 27th International Conference on Auditory Display (ICAD 2022) 24-27 June 2022, Virtual conference.This paper presents an approach to sonifying 2D cellular data. Its primary goal is attaining listener comprehension parity between original visual data and its sonified counterpart for the purpose of understanding cell behavior, including movement, mitosis (or division), and cell death. Here, we present the initial findings of the automated sonification prototype named “Cellular Stethoscope” that was assessed through a 19-subject pilot study to assess its ability to accurately reflect the cell behavior captured in the video footage. The resulting system is envisioned to serve as a foundation for a complementing and potentially more efficient approach to studying cell behavior when subjected to various pharmaceutical interventions
A Conversation on Interdisciplinary Collaboration
An engaging panel conversation on interdisciplinary collaboratio
Developing Coalitions: Computing Organizational Potential
Per Kotter’s eight-step model for leading change, there is a need for a guiding coalition, which is a social network of agents for change. The proposed presentation discusses a computational approach to describing the potential of a network’s structural position for organizational influence. A simulation of an exemplary index capable of such analysis is presented by the combination of widely accepted social network measures. RStudio was used for the necessary calculations and the social network visualizations
Aegis audio engine: integrating a real-time analog signal processing, pattern recognition, and a procedural soundtrack in a live twelve-perfomer spectacle with crowd participation
Presented at the 21st International Conference on Auditory Display (ICAD2015), July 6-10, 2015, Graz, Styria, Austria.In the following paper we present Aegis: a procedural
networked soundtrack engine driven by real-time analog
signal analysis and pattern recognition. Aegis was originally
conceived as part of Drummer Game, a game-performance-spectacle hybrid research project focusing on the depiction of a battle portrayed using terracotta soldiers. In it, each of
the twelve cohorts—divided into two armies of six—are led
by a drummer-performer who issues commands by
accurately drumming precomposed rhythmic patterns on an
original Chinese war drum. The ensuing spectacle is
envisioned to also accommodate large audience participation
whose input determines the morale of the two armies. An
analog signal analyzer utilizes efficient pattern recognition
to decipher the desired action and feed it both into the game
and the soundtrack engine. The soundtrack engine then uses
this action, as well as messages from the gaming simulation,
to determine the most appropriate soundtrack parameters
while ensuring minimal repetition and seamless transitions
between various clips that account for tempo, meter, and key
changes. The ensuing simulation offers a comprehensive
system for pattern-driven input, holistic situation
assessment, and a soundtrack engine that aims to generate a
seamless musical experience without having to resort to
cross-fades and other simplistic transitions that tend to disrupt a soundtrack’s continuity
Reimagining human capacity for location-aware aural pattern recognition: A case for immersive exocentric sonification
The following paper presents a cross-disciplinary snapshot of
21st century research in sonification and leverages the review
to identify a new immersive exocentric approach to studying
human capacity to perceive spatial aural cues. The paper
further defines immersive exocentric sonification, highlights
its unique affordances, and presents an argument for its
potential to fundamentally change the way we understand
and study the human capacity for location-aware audio
pattern recognition. Finally, the paper describes an example
of an externally funded research project that aims to tackle
this newfound research whitespace
Studies in spatial aural perception: establishing foundations for immersive sonification
Presented at the 25th International Conference on Auditory Display (ICAD 2019) 23-27 June 2019, Northumbria University, Newcastle upon Tyne, UK.The Spatial Audio Data Immersive Experience (SADIE) project aims to identify new foundational relationships pertaining to human spatial aural perception, and to validate existing relationships. Our infrastructure consists of an intuitive interaction interface, an immersive exocentric sonification environment, and a layer-based amplitude-panning algorithm. Here we highlight the systemďľ’s unique capabilities and provide findings from an initial externally funded study that focuses on the assessment of human aural spatial perception capacity. When compared to the existing body of literature focusing on egocentric spatial perception, our data show that an immersive exocentric environment enhances spatial perception, and that the physical implementation using high density loudspeaker arrays enables significantly improved spatial perception accuracy relative to the egocentric and virtual binaural approaches. The preliminary observations suggest that human spatial aural perception capacity in real-world-like immersive exocentric environments that allow for head and body movement is significantly greater than in egocentric scenarios where head and body movement is restricted. Therefore, in the design of immersive auditory displays, the use of immersive exocentric environments is advised. Further, our data identify a significant gap between physical and virtual human spatial aural perception accuracy, which suggests that further development of virtual aural immersion may be necessary before such an approach may be seen as a viable alternative
NIMEhub: Toward a Repository for Sharing and Archiving Instrument Designs
This workshop will explore the potential creation of a community database of digital musical instrument (DMI) designs. In other research communities, reproducible research practices are common, including open-source software, open datasets, established evaluation methods and community standards for research practice. NIME could benefit from similar practices, both to share ideas amongst geographically distant researchers and to maintain instrument designs after their first performances. However, the needs of NIME are different from other communities on account of NIME’s reliance on custom hardware designs and the interdependence of technology and arts practice. This half-day workshop will promote a community discussion of the potential benefits and challenges of a DMI repository and plan concrete steps toward its implementation